VQ VAE


Vector-quantized variational autoencoder (VQ VAE) is a generative model that uses vector quantization to learn discrete latent representations.

SIMART: Decomposing Monolithic Meshes into Sim-ready Articulated Assets via MLLM

Add code
Mar 24, 2026
Viaarxiv icon

Perceptio: Perception Enhanced Vision Language Models via Spatial Token Generation

Add code
Mar 19, 2026
Viaarxiv icon

Architecture-Agnostic Feature Synergy for Universal Defense Against Heterogeneous Generative Threats

Add code
Mar 16, 2026
Viaarxiv icon

High-Fidelity Text-to-Image Generation from Pre-Trained Vision-Language Models via Distribution-Conditioned Diffusion Decoding

Add code
Mar 11, 2026
Viaarxiv icon

Beyond Short-Horizon: VQ-Memory for Robust Long-Horizon Manipulation in Non-Markovian Simulation Benchmarks

Add code
Mar 10, 2026
Viaarxiv icon

$Δ$VLA: Prior-Guided Vision-Language-Action Models via World Knowledge Variation

Add code
Mar 09, 2026
Viaarxiv icon

DisQ-HNet: A Disentangled Quantized Half-UNet for Interpretable Multimodal Image Synthesis Applications to Tau-PET Synthesis from T1 and FLAIR MRI

Add code
Feb 26, 2026
Viaarxiv icon

Autoregressive Visual Decoding from EEG Signals

Add code
Feb 26, 2026
Viaarxiv icon

Humanizing Robot Gaze Shifts: A Framework for Natural Gaze Shifts in Humanoid Robots

Add code
Feb 25, 2026
Viaarxiv icon

SOM-VQ: Topology-Aware Tokenization for Interactive Generative Models

Add code
Feb 24, 2026
Viaarxiv icon